Human Activity Recognition (HAR) is an emerging technology with several applications in surveillance, security, and healthcare sectors. Noninvasive HAR systems based on Wi-Fi Channel State Information (CSI) signals can be developed leveraging the quick growth of ubiquitous Wi-Fi technologies, and the correlation between CSI dynamics and body motions. In this paper, we propose Principal Component-based Wavelet Convolutional Neural Network (or PCWCNN) -- a novel approach that offers robustness and efficiency for practical real-time applications. Our proposed method incorporates two efficient preprocessing algorithms -- the Principal Component Analysis (PCA) and the Discrete Wavelet Transform (DWT). We employ an adaptive activity segmentation algorithm that is accurate and computationally light. Additionally, we used the Wavelet CNN for classification, which is a deep convolutional network analogous to the well-studied ResNet and DenseNet networks. We empirically show that our proposed PCWCNN model performs very well on a real dataset, outperforming existing approaches.
translated by 谷歌翻译
孟加拉语是世界上说话最多的语言之一,全球有超过3亿的演讲者。尽管它很受欢迎,但由于缺乏多样化的开源数据集,对孟加拉语音识别系统的发展的研究受到阻碍。作为前进的道路,我们已经众包孟加拉语音语音数据集,这是句子级自动语音识别语料库。该数据集于Mozilla Common Voice平台上收集,是正在进行的广告系列的一部分,该活动已在2个月内收集了超过400个小时的数据,并且正在迅速增长。我们的分析表明,与OpenSLR孟加拉ASR数据集相比,该数据集具有更多的发言人,音素和环境多样性,这是最大的现有开源孟加拉语语音数据集。我们提供从数据集获得的见解,并讨论未来版本中需要解决的关键语言挑战。此外,我们报告了一些自动语音识别(ASR)算法的当前性能,并为将来的研究设定了基准。
translated by 谷歌翻译
基于Centroid的聚类方法,例如K-均值,K-Medoids和K-Centers在探索性数据分析中被大量应用作为首选工具。在许多情况下,这些方法用于获得数据歧管的代表性质心,以可视化或摘要数据集。现实世界的数据集通常包含固有的异常情况,例如重复样本和采样偏见,表现出不平衡的聚类。我们建议通过对质心形成的群集引入最大半径约束$ r $来纠正这种情况,即,从同一集群中的样本则不应以$ \ ell_2 $距离的价格分开超过$ 2R $。我们通过求解半明确程序来实现此约束,然后是二次约束的线性分配问题。通过定性结果,我们表明我们提出的方法对数据集的不平衡和采样伪像是可靠的。据我们所知,我们的是第一个受到严格半径约束的约束K-均值聚类方法。 https://bit.ly/kmeans限制的代码
translated by 谷歌翻译
制造中的一个自主实验平台据说能够进行顺序搜索,以便自行为先进材料寻找合适的制造条件,甚至用于发现具有最小的人为干预的新材料。这种平台的智能控制的核心是政策指导顺序实验,即根据到目前为止所做的事情来决定在下次进行下一个实验的地方。此类政策不可避免地违反勘探,而目前的做法是利用预期改进标准或其变体的贝叶斯优化框架。我们讨论是否利用与直接观察相关的元素和惊喜程度来促进剥削与勘探有益。我们使用两个现有的惊喜指标设计了一个惊喜的反应政策,称为香农惊喜和贝叶斯惊喜。我们的分析表明,令人惊讶的反应政策似乎更适合于在资源限制下快速表征响应面或设计地点的整体景观。我们认为未来派自治实验平台需要这种能力。我们没有声称我们有一个完全自主的实验平台,但相信我们目前的努力揭示了新灯或提供了不同的视角,因为研究人员正在赛车提升各种原始自治实验系统的自主权。
translated by 谷歌翻译
在动态威胁环境中实时跟踪多个移动对象是国家安全和监视系统中的重要因素。它有助于查明和区分潜在的候选人与其他正常物体的构成威胁,并监视异常轨迹,直到干预为止。为了找到异常运动模式,需要具有准确的数据关联算法,该算法可以将位置和运动的顺序观察与基础移动对象联系起来,因此,随着对象的移动,构建对象的轨迹。在这项工作中,我们开发了一种时空方法,用于跟踪海事血管,因为该容器的位置和运动观测是通过自动识别系统收集的。提出的方法是为了应对数据关联挑战的一种努力,在该挑战中,有意抑制船只和船舶识别的数量,并在数据集中创建时间差距,以模仿威胁环境下现实生活中的运营复杂性。在挑战中提供了三个培训数据集和五个测试集,并且数据挑战组织者设计了一组定量性能指标,用于评估和比较参与者开发的结果方法。当我们提出的赛道关联算法应用于五个测试集时,该算法得分非常具竞争力。
translated by 谷歌翻译
The performance of the Deep Learning (DL) models depends on the quality of labels. In some areas, the involvement of human annotators may lead to noise in the data. When these corrupted labels are blindly regarded as the ground truth (GT), DL models suffer from performance deficiency. This paper presents a method that aims to learn a confident model in the presence of noisy labels. This is done in conjunction with estimating the uncertainty of multiple annotators. We robustly estimate the predictions given only the noisy labels by adding entropy or information-based regularizer to the classifier network. We conduct our experiments on a noisy version of MNIST, CIFAR-10, and FMNIST datasets. Our empirical results demonstrate the robustness of our method as it outperforms or performs comparably to other state-of-the-art (SOTA) methods. In addition, we evaluated the proposed method on the curated dataset, where the noise type and level of various annotators depend on the input image style. We show that our approach performs well and is adept at learning annotators' confusion. Moreover, we demonstrate how our model is more confident in predicting GT than other baselines. Finally, we assess our approach for segmentation problem and showcase its effectiveness with experiments.
translated by 谷歌翻译
Recent advances in upper limb prostheses have led to significant improvements in the number of movements provided by the robotic limb. However, the method for controlling multiple degrees of freedom via user-generated signals remains challenging. To address this issue, various machine learning controllers have been developed to better predict movement intent. As these controllers become more intelligent and take on more autonomy in the system, the traditional approach of representing the human-machine interface as a human controlling a tool becomes limiting. One possible approach to improve the understanding of these interfaces is to model them as collaborative, multi-agent systems through the lens of joint action. The field of joint action has been commonly applied to two human partners who are trying to work jointly together to achieve a task, such as singing or moving a table together, by effecting coordinated change in their shared environment. In this work, we compare different prosthesis controllers (proportional electromyography with sequential switching, pattern recognition, and adaptive switching) in terms of how they present the hallmarks of joint action. The results of the comparison lead to a new perspective for understanding how existing myoelectric systems relate to each other, along with recommendations for how to improve these systems by increasing the collaborative communication between each partner.
translated by 谷歌翻译
Nowadays, the current neural network models of dialogue generation(chatbots) show great promise for generating answers for chatty agents. But they are short-sighted in that they predict utterances one at a time while disregarding their impact on future outcomes. Modelling a dialogue's future direction is critical for generating coherent, interesting dialogues, a need that has led traditional NLP dialogue models that rely on reinforcement learning. In this article, we explain how to combine these objectives by using deep reinforcement learning to predict future rewards in chatbot dialogue. The model simulates conversations between two virtual agents, with policy gradient methods used to reward sequences that exhibit three useful conversational characteristics: the flow of informality, coherence, and simplicity of response (related to forward-looking function). We assess our model based on its diversity, length, and complexity with regard to humans. In dialogue simulation, evaluations demonstrated that the proposed model generates more interactive responses and encourages a more sustained successful conversation. This work commemorates a preliminary step toward developing a neural conversational model based on the long-term success of dialogues.
translated by 谷歌翻译
In this work, we introduce a hypergraph representation learning framework called Hypergraph Neural Networks (HNN) that jointly learns hyperedge embeddings along with a set of hyperedge-dependent embeddings for each node in the hypergraph. HNN derives multiple embeddings per node in the hypergraph where each embedding for a node is dependent on a specific hyperedge of that node. Notably, HNN is accurate, data-efficient, flexible with many interchangeable components, and useful for a wide range of hypergraph learning tasks. We evaluate the effectiveness of the HNN framework for hyperedge prediction and hypergraph node classification. We find that HNN achieves an overall mean gain of 7.72% and 11.37% across all baseline models and graphs for hyperedge prediction and hypergraph node classification, respectively.
translated by 谷歌翻译
A "heart attack" or myocardial infarction (MI), occurs when an artery supplying blood to the heart is abruptly occluded. The "gold standard" method for imaging MI is Cardiovascular Magnetic Resonance Imaging (MRI), with intravenously administered gadolinium-based contrast (late gadolinium enhancement). However, no "gold standard" fully automated method for the quantification of MI exists. In this work, we propose an end-to-end fully automatic system (MyI-Net) for the detection and quantification of MI in MRI images. This has the potential to reduce the uncertainty due to the technical variability across labs and inherent problems of the data and labels. Our system consists of four processing stages designed to maintain the flow of information across scales. First, features from raw MRI images are generated using feature extractors built on ResNet and MoblieNet architectures. This is followed by the Atrous Spatial Pyramid Pooling (ASPP) to produce spatial information at different scales to preserve more image context. High-level features from ASPP and initial low-level features are concatenated at the third stage and then passed to the fourth stage where spatial information is recovered via up-sampling to produce final image segmentation output into: i) background, ii) heart muscle, iii) blood and iv) scar areas. New models were compared with state-of-art models and manual quantification. Our models showed favorable performance in global segmentation and scar tissue detection relative to state-of-the-art work, including a four-fold better performance in matching scar pixels to contours produced by clinicians.
translated by 谷歌翻译